This is a brief summary of paper for me to study and arrange it, CHARAGRAM: Embedding Words and Sentences via Character n-grams (Wieting et al., EMNLP 2016) I read and studied.

This paper propose character embedding for word, sentence similarity and tagging problem.

character n-gram match exactly for its vector in character n-gram embedding model.

Also they said the character n-gram captures aspects of word order and word co-occurence in CHARACGRAM-PHRASE.

They use a white-space as a special token which they think it is regarded as signal of the beggining and end of a word in CHARAGRAM-PHRASE.

In other words, they model a character-based textual sequence by \(x\) = \(<x_1, x_2, … , x_m>\), which includes space characters between words as well as special start-of_sequence and end-of_sequence characters.

The CHARAGRAM model embeds a character sequence \(x\) by adding the vectors of it character n-grams followed by an elemenwise nonlinearity.

Reference